if (size == 0)
size = PAGE_SIZE;
+ if (size == 0)
+ printk(XENLOG_WARNING "ioremap(): Trying to map %lx, size 0\n", mpaddr);
+
end = PAGE_ALIGN(mpaddr + size);
if (!iomem_access_permitted(d, mpaddr >> PAGE_SHIFT,
(end >> PAGE_SHIFT) - 1))
return -EPERM;
- return assign_domain_mmio_page(d, mpaddr, size);
+ return assign_domain_mmio_page(d, mpaddr, mpaddr, size,
+ ASSIGN_writable | ASSIGN_nocache);
}
unsigned long
u64 start = md->phys_addr;
u64 size = md->num_pages << EFI_PAGE_SHIFT;
u64 end = start + size;
+ u64 mpaddr;
unsigned long flags;
switch (md->type) {
break;
case EFI_MEMORY_MAPPED_IO_PORT_SPACE:
+ flags = ASSIGN_writable; /* dummy - zero */
+ if (md->attribute & EFI_MEMORY_UC)
+ flags |= ASSIGN_nocache;
+
+ if (start > 0x1ffffffff0000000UL) {
+ mpaddr = 0x4000000000000UL - size;
+ printk(XENLOG_INFO "Remapping IO ports from "
+ "%lx to %lx\n", start, mpaddr);
+ } else
+ mpaddr = start;
+
/* Map into dom0. */
- assign_domain_mmio_page(d, start, size);
+ assign_domain_mmio_page(d, mpaddr, start, size, flags);
/* Copy descriptor. */
*dom_md = *md;
+ dom_md->phys_addr = mpaddr;
dom_md->virt_addr = 0;
num_mds++;
break;
if (domain_page_mapped(d, addr))
continue;
- if (efi_mmio(addr, PAGE_SIZE))
- assign_domain_mmio_page(d, addr, PAGE_SIZE);
+ if (efi_mmio(addr, PAGE_SIZE)) {
+ unsigned long flags;
+ flags = ASSIGN_writable | ASSIGN_nocache;
+ assign_domain_mmio_page(d, addr, addr,
+ PAGE_SIZE, flags);
+ }
}
return num_mds;
}
}
unsigned long
-assign_domain_mmio_page(struct domain *d,
- unsigned long mpaddr, unsigned long size)
+assign_domain_mmio_page(struct domain *d, unsigned long mpaddr,
+ unsigned long phys_addr, unsigned long size,
+ unsigned long flags)
{
+ unsigned long end = PAGE_ALIGN(mpaddr + size);
if (size == 0) {
gdprintk(XENLOG_INFO, "%s: domain %p mpaddr 0x%lx size = 0x%lx\n",
__func__, d, mpaddr, size);
#endif
return -EINVAL;
}
- assign_domain_same_page(d, mpaddr, size, ASSIGN_writable | ASSIGN_nocache);
+
+ for (mpaddr &= PAGE_MASK; mpaddr < end;
+ mpaddr += PAGE_SIZE, phys_addr += PAGE_SIZE) {
+ __assign_domain_page(d, mpaddr, phys_addr, flags);
+ }
+
return mpaddr;
}
extern unsigned long lookup_domain_mpa(struct domain *d, unsigned long mpaddr, struct p2m_entry* entry);
extern void *domain_mpa_to_imva(struct domain *d, unsigned long mpaddr);
extern volatile pte_t *lookup_noalloc_domain_pte(struct domain* d, unsigned long mpaddr);
-extern unsigned long assign_domain_mmio_page(struct domain *d, unsigned long mpaddr, unsigned long size);
+extern unsigned long assign_domain_mmio_page(struct domain *d, unsigned long mpaddr, unsigned long phys_addr, unsigned long size, unsigned long flags);
extern unsigned long assign_domain_mach_page(struct domain *d, unsigned long mpaddr, unsigned long size, unsigned long flags);
int domain_page_mapped(struct domain *d, unsigned long mpaddr);
int efi_mmio(unsigned long physaddr, unsigned long size);